Simple Bayesian Classifiers Do Not Assume Independence

نویسندگان

  • Pedro M. Domingos
  • Michael J. Pazzani
چکیده

Bayes’ theorem tells us how to optimally predict the class of a previously unseen example, given a training sample. The chosen class should be the one which maximizes P(CilE) = P(Ci) P(EICi) /P(E), where Ci is the ith class, E is the test example, P(YIX) denotes the conditional probability of Y given X, and probabilities are estimated from the training sample. Let an example be a vector of a attributes. If the attributes are independent given the class, P(EICi) can be decomposed into the product P(vlICi) . . . P(va(Ci), where v.j is the value of the jth attribute in the example E. Therefore we should predict the class that maximizes:

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Theoretical and Experimental Evaluation of Augmented Bayesian Classifiers

Naive Bayes is a simple Bayesian network classifier with strong independence assumptions among features. This classifier despite its strong independence assumptions, often performs well in practice. It is believed that relaxing the independence assumptions of naive Bayes may improve the performance of the resulting structure. Augmented Bayesian Classifiers relax the independence assumptions of ...

متن کامل

Beyond Independence: Conditions for the Optimality of the Simple Bayesian Classi er

The simple Bayesian classi er (SBC) is commonly thought to assume that attributes are independent given the class, but this is apparently contradicted by the surprisingly good performance it exhibits in many domains that contain clear attribute dependences. No explanation for this has been proposed so far. In this paper we show that the SBC does not in fact assume attribute independence, and ca...

متن کامل

Beyond Independence: Conditions for the Optimality of the Simple Bayesian Classifier

The simple Bayesian classi er (SBC) is commonly thought to assume that attributes are independent given the class, but this is apparently contradicted by the surprisingly good performance it exhibits in many domains that contain clear attribute dependences. No explanation for this has been proposed so far. In this paper we show that the SBC does not in fact assume attribute independence, and ca...

متن کامل

Tractable Bayesian Learning of Tree Augmented Naive Bayes Classifiers

Bayesian classifiers such as Naive Bayes or Tree Augmented Naive Bayes (TAN) have shown excellent performance given their simplicity and heavy underlying independence assumptions. In this paper we introduce a classifier taking as basis the TAN models and taking into account uncertainty in model selection. To do this we introduce decomposable distributions over TANs and show that the expression ...

متن کامل

Sisterhood of Classifiers: A Comparative Study of Naive Bayes and Noisy-or Networks

Classification is a task central to many machine learning problems. In this paper we examine two Bayesian network classifiers, the naive Bayes and the noisy-or models. They are of particular interest because of their simple structures. We compare them on two dimensions: expressive power and ability to learn. As it turns out, naive Bayes, noisy-or, and logistic regression classifiers all have eq...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1996